Communication and matrix computations on large message passing systems

نویسنده

  • G. W. Stewart
چکیده

abstract This paper is concerned with the consequences for matrix computations of having a rather large number of general purpose processors, say ten or twenty thousand, connected in a network in such a way that a processor can communicate only with its immediate neighbors. Certain communication tasks associated with most matrix algorithms are deened and formulas developed for the time required to perform them under several communication regimes. The results are compared with the times for a nominal n 3 oating point operations. The results suggest that it is possible to use a large number of processors to solve matrix problems at a relatively ne granularity, provided ne grain communication is available.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Communication and Matrix Computations on Large Message Passing Systems Communication and Matrix Computations on Large Message Passing Systems

abstract This paper is concerned with the consequences for matrix computations of having a rather large number of general purpose processors, say ten or twenty thousand, connected in a network in such a way that a processor can communicate only with its immediate neighbors. Certain communication tasks associated with most matrix algorithms are deened and formulas developed for the time required...

متن کامل

A Message-Passing Distributed Memory Parallel Algorithm for a Dual-Code Thin Layer, Parabolized Navier-Stokes Solver

In this study, the results of parallelization of a 3-D dual code (Thin Layer, Parabolized Navier-Stokes solver) for solving supersonic turbulent flow around body and wing-body combinations are presented. As a serial code, TLNS solver is very time consuming and takes a large part of memory due to the iterative and lengthy computations. Also for complicated geometries, an exceeding number of grid...

متن کامل

Parallel Reduction from Block Hessenberg to Hessenberg using MPI

In many scientific applications, eigenvalues of a matrix have to be computed. By first reducing a matrix from fully dense to Hessenberg form, eigenvalue computations with the QR algorithm become more efficient. Previous work on shared memory architectures has shown that the Hessenberg reduction is in some cases most efficient when performed in two stages: First reduce the matrix to block Hessen...

متن کامل

Performance Characterization of Matrix Multiplication on SGI Altix 3700

Matrix multiplication is widely used in a variety of applications and is often one of the core components of many scientific computations which includes graph theory, numerical methods, digital control and signal processing. Multiplication of large matrices require a lot of computation time as its complexity is O(n), where n is the dimension of the matrix. A serial algorithm to compute large ma...

متن کامل

Compiling P–GRADE programs for the JGrid architecture∗

Grid systems are large, geographically distributed computational environments. JGrid is a Grid infrastructure, that provides a software environment for such systems. It is built on top of the Java-based Jini technology, which provides a common platform and a service-based lookup system. P–GRADE is an interactive, graphical software development tool, designed to build distributed parallel progra...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:
  • Parallel Computing

دوره 16  شماره 

صفحات  -

تاریخ انتشار 1990